# Long-context reasoning
Deepseek R1 0528 AWQ
MIT
AWQ-quantized version of DeepSeek R1 0528, supports full-context-length operation on 8x80GB GPUs using vLLM.
Large Language Model
Transformers Supports Multiple Languages

D
cognitivecomputations
145
7
Qwenlong L1 32B
Apache-2.0
QwenLong-L1 is a long-context reasoning model trained with reinforcement learning, demonstrating excellent performance across seven long-context document QA benchmarks.
Large Language Model
Transformers

Q
Tongyi-Zhiwen
683
106
Llama 3.1 Nemotron Nano 4B V1.1 GGUF
Other
A 4B-parameter large language model released by NVIDIA, supporting 128k tokens context length, optimized for reasoning, dialogue, and RAG tasks
Large Language Model English
L
lmstudio-community
588
1
AM Thinking V1 GGUF
Apache-2.0
AM Thinking v1 is a large language model developed by the A-M team based on Qwen 2.5-32B-Base, with enhanced reasoning capabilities and support for a context length of 132k tokens.
Large Language Model
A
lmstudio-community
306
1
M1ndb0t 0M3N Q4 K M GGUF
High-performance GGUF conversion version based on Qwen3-14B large language model, optimized for creative reasoning, deep dream logic, agent interaction, and multilingual instructions
Large Language Model English
M
TheMindExpansionNetwork
18
1
Qwen3 4B NEO Imatrix Max GGUF
Apache-2.0
This is a NEO Imatrix quantized version based on the Qwen3-4B model, using BF16 format MAX output tensors to enhance reasoning and output generation capabilities, supporting 32k context length.
Large Language Model
Q
DavidAU
1,152
3
Delta Pavonis Qwen 14B
Apache-2.0
Enhanced reasoning model based on Qwen-2.5 14B architecture, optimized for general reasoning and Q&A scenarios, supporting 128K context and 8K output
Large Language Model
Transformers

D
prithivMLmods
547
3
Llama 3 70b Arimas Story RP V1.6 4.0bpw H6 Exl2
A merged model based on Llama 3 70B architecture, optimized for story generation and role-play, supporting long context windows
Large Language Model
Transformers

L
kim512
20
1
Granite 3.2 2b Instruct GGUF
Apache-2.0
Granite-3.2-2B-Instruct is a 2-billion-parameter long-context AI model specifically fine-tuned for reasoning capabilities. Built upon Granite-3.1-2B-Instruct, it was trained using a mix of permissively licensed open-source datasets and internally generated synthetic data to enhance performance on reasoning tasks.
Large Language Model
G
Mungert
754
3
Theta Lyrae Qwen 14B
Apache-2.0
Theta-Lyrae-Qwen-14B is a 14-billion-parameter model based on the Qwen 2.5 14B modal architecture, optimized for general reasoning and Q&A capabilities, excelling in context understanding, logical reasoning, and multi-step problem-solving.
Large Language Model
Transformers

T
prithivMLmods
21
2
Galactic Qwen 14B Exp2
Apache-2.0
Galactic-Qwen-14B-Exp2 is a large language model based on the Qwen 2.5 14B architecture, focusing on enhanced reasoning capabilities, excelling in context understanding, logical reasoning, and multi-step problem solving.
Large Language Model
Transformers Supports Multiple Languages

G
prithivMLmods
558
4
Romboultima 32B
MIT
RombUltima-32B is a fusion model that combines the strengths of Rombos-LLM-V2.5-Qwen-32b and Ultima-32B, optimizing reasoning capabilities, multilingual understanding, and multi-turn dialogue performance.
Large Language Model
Transformers

R
FINGU-AI
75
4
Deepseek R1 AWQ
MIT
AWQ quantized version of DeepSeek R1 model, optimized for float16 overflow issues and supports efficient inference deployment
Large Language Model
Transformers Supports Multiple Languages

D
cognitivecomputations
30.46k
77
Modernbert Base Nli
Apache-2.0
ModernBERT is a model fine-tuned on multi-task natural language inference (NLI) tasks, excelling in zero-shot classification and long-context reasoning.
Large Language Model
Transformers Supports Multiple Languages

M
tasksource
1,867
20
Phi 3 Small 128k Instruct
MIT
Phi-3-Small-128K-Instruct is a 7-billion-parameter lightweight open-source model focused on high quality and strong reasoning capabilities, supporting 128K long context, and excelling in tasks such as commonsense reasoning, language understanding, mathematics, and coding.
Large Language Model
Transformers Other

P
microsoft
7,194
176
Phi 3 Medium 128k Instruct
MIT
Phi-3-Medium-128K-Instruct is a lightweight open-source model with 14 billion parameters, focusing on high quality and strong reasoning capabilities, supporting a 128K context length.
Large Language Model
Transformers Other

P
microsoft
17.52k
381
C4ai Command R Plus Imat.gguf
C4AI Command R+ is a 104B parameter multilingual large language model supporting Retrieval-Augmented Generation (RAG) and tool calling, optimized for tasks like reasoning, summarization, and Q&A.
Large Language Model
C
dranger003
2,783
140
Einstein V4 7B
Other
Einstein-v4-7B is a large language model fully fine-tuned on diverse scientific datasets based on Mistral-7B-v0.1, specializing in STEM field tasks
Large Language Model
Transformers English

E
Weyaxi
43
49
Tinyllama 1.1B 32k
Apache-2.0
A 32k-context fine-tuned version based on TinyLlama-1.1B, achieving long-context processing capability by increasing rope theta
Large Language Model
Transformers English

T
Doctor-Shotgun
51
29
Featured Recommended AI Models